Skip to content

Conversation

@Kacper-Pietkun
Copy link
Contributor

@Kacper-Pietkun Kacper-Pietkun commented Oct 23, 2025

I added tests for custom ops defined in vllm_gaudi/ops:

  • For the tests of ops that are not using cuda kernels - native ops and hpu ops are triggered for the same input and their outputs are compared
  • For others tests that are using cuda kernels (so cannot be called with vllm-gaudi plugin) I created separate directory to store some predefined small tensors - weights, inputs and outputs. These tensors are too big to hardcode them in tests, however their sizes were adjusted, so all of them weight less than 3MB in total. Tensors are stored in a .safetensors format. Such tests run hpu ops with loaded inputs and weights and compare their outputs with the loaded outputs.

Signed-off-by: Kacper Pietkun <[email protected]>
@Kacper-Pietkun
Copy link
Contributor Author

/run-gaudi-tests

Signed-off-by: Kacper Pietkun <[email protected]>
@Kacper-Pietkun
Copy link
Contributor Author

/run-gaudi-tests

@github-actions
Copy link

✅ CI Passed

All checks passed successfully against the following vllm commit:
7e0941055fdf89bae93045683dd80542177f3241

@Kacper-Pietkun
Copy link
Contributor Author

/run-gaudi-tests

@github-actions
Copy link

✅ CI Passed

All checks passed successfully against the following vllm commit:
63b22e0dbb901b75619aa4bca2dfa1d7a71f439e

Signed-off-by: Kacper Pietkun <[email protected]>
@github-actions
Copy link

🚧 CI Blocked

The main CI workflow was not started for the following reason:

Your branch is behind the base branch. Please merge or rebase to get the latest changes.

@Kacper-Pietkun
Copy link
Contributor Author

/run-gaudi-tests

@github-actions
Copy link

✅ CI Passed

All checks passed successfully against the following vllm commit:
b368382964913312d41c670b4166f4c83eed49aa

@github-actions
Copy link

github-actions bot commented Nov 4, 2025

✅ CI Passed

All checks passed successfully against the following vllm commit:
0384aa7150c4c9778efca041ffd1beb3ad2bd694

Copilot AI review requested due to automatic review settings November 7, 2025 09:35
Copy link
Contributor

Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull Request Overview

This PR adds comprehensive unit tests for custom operators implemented in vllm_gaudi/ops. The tests verify correctness by comparing outputs between native VLLM operators and HPU-specific implementations. For operators using CUDA kernels, pre-computed reference tensors stored in safetensors format are used for validation.

Key changes:

  • Native and HPU operator outputs are compared for operators compatible with both implementations
  • Pre-computed reference data in safetensors format is used for CUDA kernel-based operators
  • Test utilities added for temporary operator registry management and test data access

Reviewed Changes

Copilot reviewed 11 out of 19 changed files in this pull request and generated 5 comments.

Show a summary per file
File Description
vllm_gaudi/ops/hpu_fp8.py Removed unused imports and decorator
tests/unit_tests/test_bucketing.py Added config clearing before setup
tests/unit_tests/ops/utils.py Added test utilities for operator registration and data loading
tests/unit_tests/ops/test_hpu_rotary_embedding.py Tests for rotary embedding operator variants
tests/unit_tests/ops/test_hpu_multihead_attn.py Tests for multi-head attention operator
tests/unit_tests/ops/test_hpu_layernorm.py Tests for RMS normalization operator
tests/unit_tests/ops/test_hpu_gptq.py Tests for GPTQ quantization operator
tests/unit_tests/ops/test_hpu_fused_moe.py Tests for fused MoE operator
tests/unit_tests/ops/test_hpu_fp8.py Tests for FP8 quantization operators
tests/unit_tests/ops/test_hpu_compressed_tensors.py Tests for compressed tensor operators
tests/unit_tests/ops/test_hpu_awq.py Tests for AWQ quantization operator

💡 Add Copilot custom instructions for smarter, more guided reviews. Learn how to get started.

Kacper-Pietkun and others added 2 commits November 7, 2025 10:41
Co-authored-by: Copilot <[email protected]>
Signed-off-by: Kacper Pietkun <[email protected]>
Co-authored-by: Copilot <[email protected]>
Signed-off-by: Kacper Pietkun <[email protected]>
Kacper-Pietkun and others added 3 commits November 7, 2025 10:42
Co-authored-by: Copilot <[email protected]>
Signed-off-by: Kacper Pietkun <[email protected]>
@Kacper-Pietkun
Copy link
Contributor Author

All of the above changes are just corrections of spelling mistakes detected by copilot

@github-actions
Copy link

github-actions bot commented Nov 7, 2025

✅ CI Passed

All checks passed successfully against the following vllm commit:
0384aa7150c4c9778efca041ffd1beb3ad2bd694

@michalkuligowski michalkuligowski merged commit 0a6113b into vllm-project:main Nov 7, 2025
37 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants